BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova
Google AI Language
Submitted on 11 Oct 2018 (v1), last revised 24 May 2019 (this version, v2)
https://arxiv.org/abs/1810.04805